The Articulated Scene Model: Model-less Priors for Robot Object Learning?
نویسندگان
چکیده
Human analysis of dynamic scenes consists of two parallel processing chains [2]. The first one concentrates on motion which is defined as variation of location while the second one processes change which is the variation of structure. The detection of a scene change is realized phenomenologically by comparing currently visible structures with a representation in memory. These psychological findings have motivated us to design an articulated scene modeling approach [1] which enables a robot to extract articulated scene parts through observing the spatial changes caused by their manipulation. This approach processes a sequence of 3D scans taken from a fixed view point which captures a dynamic scene where a human moves around and manipulates the environment by, e.g., replacing chairs or opening doors. It estimates per frame Ft the actively moving persons Et, the so far static scene background St, and movable objects / articulated scene parts Ot. The moving persons are tracked using a particle filter with a weak cylinder model. Static and movable scene parts are computed by a comparison of the current frame (where the tracked persons have been excluded) with the background model St−t estimated from the previous frames. For dense depth sensors, like the SwissRanger camera or the Kinect camera, such a comparison can be implemented as pixel-wise subtraction of St−t from Ft. Using the physical fact that per pixel the farthest static depth measurements along a ray define the static background, the background model is instantaneously adapted to newly uncovered background while arbitrary movable objects (like a replaced chair or an opened cupboard door) arise model-less from depth measurements emerging in front of the known static background. The video 1 2 shows for an Swissranger sequence the emerging of the static background (in blue), the movable objects (in orange), and the trajectories of an entity (in cyan and green) for two view points. The scene modeling part of our approach can also be presented on site in real-time on Kinect data. The development of cameras like the Kinect camera which combine dense depth measurement with a normal color camera in an elegant way, opens up new possibilities for interactive object learning. Future work could concentrate on the question whether the extracted movable objects (like chair) can be used to compute suitable features that can be used to detect, for example, other chairs in the scene which have not been moved, so far. Further, a history of positions of an articulated object like a drawer provides model-less tracks of object parts which can be used to train candidate kinematic models (like rotational, rigid, prismatic) for the observed tracks [3]. kinematic model Prior for vision-based detection?
منابع مشابه
Using Articulated Scene Models for Dynamic 3D Scene Analysis in Vista Spaces
In this paper we describe an efficient but detailed new approach to analyze complex dynamic scenes directly in 3D. The arising information is important for mobile robots to solve tasks in the area of household robotics. In our work a mobile robot builds an articulated scene model by observing the environment in the visual field or rather in the so-called vista space. The articulated scene model...
متن کاملVisual Tracking using Learning Histogram of Oriented Gradients by SVM on Mobile Robot
The intelligence of a mobile robot is highly dependent on its vision. The main objective of an intelligent mobile robot is in its ability to the online image processing, object detection, and especially visual tracking which is a complex task in stochastic environments. Tracking algorithms suffer from sequence challenges such as illumination variation, occlusion, and background clutter, so an a...
متن کاملAdaptive RBF network control for robot manipulators
TThe uncertainty estimation and compensation are challenging problems for the robust control of robot manipulators which are complex systems. This paper presents a novel decentralized model-free robust controller for electrically driven robot manipulators. As a novelty, the proposed controller employs a simple Gaussian Radial-Basis-Function Network as an uncertainty estimator. The proposed netw...
متن کاملSpatiotemporal Articulated Models for Dynamic SLAM
We propose an online spatiotemporal articulation model estimation framework that estimates both articulated structure as well as a temporal prediction model solely using passive observations. The resulting model can predict future motions of an articulated object with high confidence because of the spatial and temporal structure. We demonstrate the effectiveness of the predictive model by incor...
متن کاملArticulated Motion Learning via Visual and Lingual Signals
In order for robots to operate effectively in homes and workplaces, they must be able to manipulate the articulated objects common to environments built for and by humans. Previous work learns kinematic models that prescribe this manipulation from visual demonstrations. Lingual signals, such as natural language descriptions and instructions, offer a complementary means of conveying knowledge of...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2011